MLA-LSTM: A Local and Global Location Attention LSTM Learning Model for Scoring Figure Skating
نویسندگان
چکیده
Video-based scoring using neural networks is a very important means for evaluating many sports, especially figure skating. Although methods action quality have been proposed, there no uniform conclusion on the best feature extractor and clip length existing methods. Furthermore, during aggregation stage, these cannot accurately locate target information. To address tasks, firstly, we systematically compare effects of skating model with three different extractors (C3D, I3D, R3D) four segment lengths (5, 8, 16, 32). Secondly, propose Multi-Scale Location Attention Module (MS-LAM) to capture location information athletes in video frames. Finally, present novel Multi-scale Attentive Long Short-Term Memory (MLA-LSTM), which can efficiently learn local global sequence each video. In addition, our proposed has validated Fis-V MIT-Skate datasets. The experimental results show that I3D 32 frames per second are tasks. outperforms current state-of-the-art method hybrid dynAmic-statiC conText-aware attentION NETwork (ACTION-NET), (by 0.069 Spearman’s rank correlation). it achieves average improvements 0.059 compared convolutional skip Self-attentive LSTM (MS-LSTM). It demonstrates effectiveness models learning score videos.
منابع مشابه
Future Context Attention for Unidirectional LSTM Based Acoustic Model
Recently, feedforward sequential memory networks (FSMN) has shown strong ability to model past and future long-term dependency in speech signals without using recurrent feedback, and has achieved better performance than BLSTM in acoustic modeling. However, the encoding coefficients in FSMN is context-independent while context-dependent weights are commonly supposed to be more reasonable in acou...
متن کاملSimulate Congestion Prediction in a Wireless Network Using the LSTM Deep Learning Model
Achieved wireless networks since its beginning the prevalent wide due to the increasing wireless devices represented by smart phones and laptop, and the proliferation of networks coincides with the high speed and ease of use of the Internet and enjoy the delivery of various data such as video clips and games. Here's the show the congestion problem arises and represent aim of the research is t...
متن کاملA Hierarchical LSTM Model for Joint Tasks
Previous work has shown that joint modeling of two Natural Language Processing (NLP) tasks are effective for achieving better performances for both tasks. Lots of task-specific joint models are proposed. This paper proposes a Hierarchical Long Short-Term Memory (HLSTM) model and some its variants for modeling two tasks jointly. The models are flexible for modeling different types of combination...
متن کاملHashtag Recommendation with Topical Attention-Based LSTM
Microblogging services allow users to create hashtags to categorize their posts. In recent years, the task of recommending hashtags for microblogs has been given increasing attention. However, most of existing methods depend on hand-crafted features. Motivated by the successful use of long short-term memory (LSTM) for many natural language processing tasks, in this paper, we adopt LSTM to learn...
متن کاملLearning Natural Language Inference using Bidirectional LSTM model and Inner-Attention
In this paper, we proposed a sentence encoding-based model for recognizing text entailment. In our approach, the encoding of sentence is a two-stage process. Firstly, average pooling was used over word-level bidirectional LSTM (biLSTM) to generate a firststage sentence representation. Secondly, attention mechanism was employed to replace average pooling on the same sentence for better represent...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Systems
سال: 2023
ISSN: ['2079-8954']
DOI: https://doi.org/10.3390/systems11010021